Distributed Stochastic Nested Optimization for Emerging Machine Learning Models: Algorithm and Theory
نویسندگان
چکیده
Traditional machine learning models can be formulated as the expected risk minimization (ERM) problem: minw∈Rd Eξ [l(w; ξ)], where w ∈ Rd denotes model parameter, ξ represents training samples, l(·) is loss function. Numerous optimization algorithms, such stochastic gradient descent (SGD), have been developed to solve ERM problem. However, a wide range of emerging are beyond this class problems, model-agnostic meta-learning (Finn, Abbeel, and Levine 2017). Of particular interest my research nested (SNO) problem, whose objective function has structure. Specifically, I focusing on two instances kind compositional (SCO) which cover meta-learning, area-under-the-precision recall-curve optimization, contrastive self-supervised learning, etc., bilevel (SBO) applied hyperparameter neural network architecture search, etc. With emergence large-scale distributed data, user data generated mobile devices or intelligent hardware, it imperative develop algorithms for SNO (Distributed SNO). A significant challenge optimizing problems lies in that (hyper-)gradient biased estimation full gradient. Thus, existing when them suffer from slow convergence rates. In talk, will discuss recent works about SCO (Gao Huang 2021; Gao, Li, 2022) SBO (Gao, Gu, Thai 2022; Gao under both centralized decentralized settings, including algorithmic details reducing bias gradient, theoretical rate, practical applications, then highlight challenges future research.
منابع مشابه
Stochastic, Distributed and Federated Optimization for Machine Learning
We study optimization algorithms for the finite sum problems frequently arising in machine learning applications. First, we propose novel variants of stochastic gradient descent with a variance reduction property that enables linear convergence for strongly convex objectives. Second, we study distributed setting, in which the data describing the optimization problem does not fit into a single c...
متن کاملA Hybrid Optimization Algorithm for Learning Deep Models
Deep learning is one of the subsets of machine learning that is widely used in Artificial Intelligence (AI) field such as natural language processing and machine vision. The learning algorithms require optimization in multiple aspects. Generally, model-based inferences need to solve an optimized problem. In deep learning, the most important problem that can be solved by optimization is neural n...
متن کاملA Hybrid Optimization Algorithm for Learning Deep Models
Deep learning is one of the subsets of machine learning that is widely used in Artificial Intelligence (AI) field such as natural language processing and machine vision. The learning algorithms require optimization in multiple aspects. Generally, model-based inferences need to solve an optimized problem. In deep learning, the most important problem that can be solved by optimization is neural n...
متن کاملStochastic Optimization for Machine Learning
It has been found that stochastic algorithms often find good solutions much more rapidly than inherently-batch approaches. Indeed, a very useful rule of thumb is that often, when solving a machine learning problem, an iterative technique which relies on performing a very large number of relatively-inexpensive updates will often outperform one which performs a smaller number of much "smarter" bu...
متن کاملStochastic Methods For Optimization and Machine Learning
In this project a stochastic method for general purpose optimization and machine learning is described. The method is derived from basic information-theoretic principles and generalizes the popular Cross Entropy method. The effectiveness of the method as a tool for statistical modeling and Monte Carlo simulation is demonstrated with an application to the problems of density estimation and data ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence
سال: 2023
ISSN: ['2159-5399', '2374-3468']
DOI: https://doi.org/10.1609/aaai.v37i13.26804